Markov Chain Monte Carlo and Gibbs Sampling
نویسنده
چکیده
A major limitation towards more widespread implementation of Bayesian approaches is that obtaining the posterior distribution often requires the integration of high-dimensional functions. This can be computationally very difficult, but several approaches short of direct integration have been proposed (reviewed by Smith 1991, Evans and Swartz 1995, Tanner 1996). We focus here on Markov Chain Monte Carlo (MCMC) methods, which attempt to simulate direct draws from some complex distribution of interest. MCMC approaches are so-named because one uses the previous sample values to randomly generate the next sample value, generating a Markov chain (as the transition probabilities between sample values are only a function of the most recent sample value). The realization in the early 1990’s (Gelfand and Smith 1990) that one particular MCMC method, the Gibbs sampler, is very widely applicable to a broad class of Bayesian problems has sparked a major increase in the application of Bayesian analysis, and this interest is likely to continue expanding for sometime to come. MCMC methods have their roots in the Metropolis algorithm (Metropolis and Ulam 1949, Metropolis et al. 1953), an attempt by physicists to compute complex integrals by expressing them as expectations for some distribution and then estimate this expectation by drawing samples from that distribution. The Gibbs sampler (Geman and Geman 1984) has its origins in image processing. It is thus somewhat ironic that the powerful machinery of MCMC methods had essentially no impact on the field of statistics until rather recently. Excellent (and detailed) treatments of MCMC methods are found in Tanner (1996) and Chapter two of Draper (2000). Additional references are given in the particular sections below.
منابع مشابه
A Theoretical and Practical Implementation Tutorial on Topic Modeling and Gibbs Sampling
This technical report provides a tutorial on the theoretical details of probabilistic topic modeling and gives practical steps on implementing topic models such as Latent Dirichlet Allocation (LDA) through the Markov Chain Monte Carlo approximate inference algorithm Gibbs Sampling.
متن کاملEnsuring Rapid Mixing and Low Bias for Asynchronous Gibbs Sampling
Gibbs sampling is a Markov chain Monte Carlo technique commonly used for estimating marginal distributions. To speed up Gibbs sampling, there has recently been interest in parallelizing it by executing asynchronously. While empirical results suggest that many models can be efficiently sampled asynchronously, traditional Markov chain analysis does not apply to the asynchronous case, and thus asy...
متن کاملMarkov Chain Monte Carlo Methods : Computation and Inference
This chapter reviews the recent developments in Markov chain Monte Carlo simulation methods These methods, which are concerned with the simulation of high dimensional probability distributions, have gained enormous prominence and revolutionized Bayesian statistics The chapter provides background on the relevant Markov chain theory and provides detailed information on the theory and practice of ...
متن کاملBayesian Analysis of the Stochastic Switching Regression Model Using Markov Chain Monte Carlo Methods
This study develops Bayesian methods of estimating the parameters of the stochastic switching regression model. Markov Chain Monte Carlo methods data augmentation and Gibbs sampling are used to facilitate estimation of the posterior means. The main feature of these two methods is that the posterior means are estimated by the ergodic averages of samples drawn from conditional distributions which...
متن کاملMarkov Chain Monte Carlo methods: Implementation and comparison
The paper and presentation will focus on MCMC methods, implemented together in MC2Pack, an ox package which allows you to run a range of sampling algorithm (MH, Gibbs, Griddy Gibbs, Adaptive Polar Importance Sampling, Adaptive Polar Sampling, and Adaptive Rejection Metropolis Sampling) on a given posterior. Computation of the marginal likelihood for the model is also done automatically, allowin...
متن کاملTheoretical rates of convergence forMarkov chain Monte
We present a general method for proving rigorous, a priori bounds on the number of iterations required to achieve convergence of Markov chain Monte Carlo. We describe bounds for spe-ciic models of the Gibbs sampler, which have been obtained from the general method. We discuss possibilities for obtaining bounds more generally.
متن کامل